Fast computation of maximum entropy / minimum divergence feature gain
نویسنده
چکیده
Maximum entropy / minimum divergence modeling is a powerful technique for constructing probability models, which has been applied to a wide variety of problems in natural language processing. A maximum entropy / minimum divergence (memd) model is built from a base model, and a set of feature functions, also called simply features, whose empirical expectations on some training corpus are known. A fundamental di culty with this technique is that while there are typically millions of features that could be incorporated into a given model, in general it is not computationally feasible, or even desirable, to use them all. Thus some means must be devised for determining each feature's predictive power, also known as its gain. Once the gains are known, the features can be ranked according to their utility, and only the most gainful ones retained. This paper presents a new algorithm for computing feature gain that is fast, accurate and memory-e cient.
منابع مشابه
A Maximum Entropy/minimum Divergence Translation Model
I present empirical comparisons between a standard statistical translation model and an equivalent Maximum Entropy/Minimum Divergence (MEMD) model, using several diierent methods for automatic feature selection. Results show that the MEMD model signiicantly outperforms the standard model in test corpus perplexity, even though it has far fewer parameters.
متن کاملA Maximum Entropy/Minimum Divergence Translation Model
I present empirical comparisons between a linear combination of standard statistical language and translation models and an equivalent Maximum Entropy/Minimum Divergence (MEMD) model, using several diierent methods for automatic feature selection. The MEMD model signiicantly outperforms the standard model in test corpus per-plexity, even though it has far fewer parameters.
متن کاملA Comparison of Criteria for Maximum Entropy/ Minimum Divergence Feature Selection
In this paper we study the gain a naturally arising statistic from the theory of memd modeling as a gure of merit for selecting features for an memd language model We compare the gain with two popular alternatives empirical activation and mutual information and argue that the gain is the preferred statistic on the grounds that it directly measures a fea ture s contribution to improving upon the...
متن کاملMinimax Entropy Principle and Its Application to Texture Modeling
This article proposes a general theory and methodology, called the minimax entropy principle, for building statistical models for images (or signals) in a variety of applications. This principle consists of two parts. The first is the maximum entropy principle for feature binding (or fusion): for a given set of observed feature statistics, a distribution can be built to bind these feature stati...
متن کاملMinimax Entropy Principle and Its Applicationto Texture
This article proposes a general theory and methodology, called the minimax entropy principle, for building statistical models for images (or signals) in a variety of applications. This principle consists of two parts. The rst is the maximum entropy principle for feature binding (or fusion): for a given set of observed feature statistics, a distribution can be built to bind these feature statist...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1998